Of the Svm - Based Runs in the High - Level Feature Extraction Task

نویسندگان

  • Mats Sjöberg
  • Ville Viitaniemi
  • Markus Koskela
  • Jorma Laaksonen
چکیده

Our experiments in TRECVID 2009 include participation in the high-level feature extraction and automatic search tasks. In the high-level feature extraction task, we used a feature fusion-based general system architecture utilizing a large number of SVM detectors, followed by a post-processing stage utilizing the concepts’ temporal and inter-concept co-occurrences. We submitted the following six runs: • PicSOM.base: Baseline run using our SOM-based HLF detection method • PicSOM.A-ngram: Baseline SVM-based run using HLF-wise geometric mean fusion and temporal n-gram post-processing • PicSOM.B-ngram: As previous, but includes also early fusion, multi-fold SFBS fusion, and more elaborate SVM training • PicSOM.E-ngram: As previous, but includes two-stage fusion utilizing cross-concept co-occurrence • PicSOM.spec-ngram: A run where the used method was selected for each HLF separately using cross-validation • PicSOM.spec-any: As previous, but the post-processing used also clustering-based inter-concept co-occurrence analysis The results show that feature fusion can consistently outperform all single features, multi-fold SFBS performed best of the tested fusion methods, and that temporal n-gram analysis is beneficial. Early fusion, and post-processing based on inter-concept co-occurrences did not improve the performance. In the search task, we concentrated on the fully-automatic runs and standard search task. We combined ASR/MT text search and concept-based retrieval. If none of the concept models could be matched with the query, we used content-based retrieval based on the video and image examples instead. We submitted the following ten fully-automatic runs: • F_A_N_PicSOM_1_10: text search baseline • F_A_N_PicSOM_2_9: visual baseline • F_A_N_PicSOM_3_8: own concepts • F_A_N_PicSOM_4_7: own concepts + text search • F_A_N_PicSOM_5_6: donated concepts • F_A_N_PicSOM_6_5: donated concepts + text search • F_A_N_PicSOM_7_4: own + donated concepts • F_A_N_PicSOM_8_3: own + donated concepts + text search • F_A_N_PicSOM_9_2: own + donated (dupl.) concepts • F_A_N_PicSOM_10_1: own + donated (dupl.) concepts + text search In the above list, “own” concepts refer to our own HLF detectors and “donated” concepts consist of MediaMill (MM) concepts + CU-VIREO374 concepts. In other than the last two runs, CU-VIREO374 are only used for words for which no MediaMill concept could be matched. The results show again that concept-based retrieval performed better than content-based search alone. Text search made a small improvement in combination with other modalities, but performed really badly on its own. Concept-selection was done both with word-matching and example-based matching, i.e. selecting concepts based on how well they would fit our own concept models.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Developing a New Method in Object Based Classification to Updating Large Scale Maps with Emphasis on Building Feature

According to the cities expansion, updating urban maps for urban planning is important and its effectiveness is depend on the information extraction / change detection accuracy. Information extraction methods are divided into two groups, including Pixel-Based (PB) and Object-Based (OB). OB analysis has overcome the limitations of PB analysis (producing salt-pepper results and features with hole...

متن کامل

UC3M High Level Feature Extraction at TRECVID 2008

This paper describes experiments carried out by the UC3M team for TRECVID 2008 high-level feature extraction task. Being our first participation in TRECVID, our goal this year has been to develop a modular system to facilitate future developments and incorporation of new functionality (feature extraction and classification modules). We have basically carried out experiments with two different k...

متن کامل

FudaSys Video Retrieval in TRECVID 2012

The video retrieval system we developed for TRECVID 2012 mainly involves the semantic indexing task which includes key frame extraction, low level feature extraction, classification and concept fusion. We extracted a new low level feature, explored various classification and fusion schemes. Four “light” runs and two 2 “pair” runs we submitted are as follows: L_A_FudaSys1: Fusion based on concep...

متن کامل

ISM TRECVID2008 High-level Feature Extraction

We studied two methods for the high-level feature extraction (HLF) task: (1) a method based on support vector machines (SVMs) with walk-based graph kernels [1], and (2) a method based on the prefixspan boosting (pboost) algorithm [2]. In the former method, each image is first segmented into a finite set of homogeneous segments and then represented as a segmentation graph where each vertex is a ...

متن کامل

Shanghai Jiao Tong University participation in high-level feature extraction, automatic search and surveillance event detectionat TRECVID 2008

In this paper, we describe our participation for high-level feature extraction, automatic search and surveillance event detection at TRECVID 2008 evaluation. In high-level feature extraction, we use selective attention model to extract visual salient feature which highlights the most visual attractive information of an image. Besides this, we extract 7 low-level features for various modalities ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2009